Latent Predictor Networks for Code Generation
نویسندگان
چکیده
Many language generation tasks require the production of text conditioned on both structured and unstructured inputs. We present a novel neural network architecture which generates an output sequence conditioned on an arbitrary number of input functions. Crucially, our approach allows both the choice of conditioning context and the granularity of generation, for example characters or tokens, to be marginalised, thus permitting scalable and effective training. Using this framework, we address the problem of generating programming code from a mixed natural language and structured specification. We create two new data sets for this paradigm derived from the collectible trading card games Magic the Gathering and Hearthstone. On these, and a third preexisting corpus, we demonstrate that marginalising multiple predictors allows our model to outperform strong benchmarks.
منابع مشابه
Optimal Combined and Adaptive Protection of Active Distribution Networks Considering Different System Topologies Incorporating Optimal Selection of Standard Relay Curves
The change in the topology of active distribution networks (ADNs) is one of the essential challenges that might affect the protection schemes. The conventional protection schemes based on base topology result in some coordination constraint violations in other topologies due to the outage of upstream substations and distributed generation units. In this article, new combinational and adaptive p...
متن کاملSyncGAN: Synchronize the Latent Space of Cross-modal Generative Adversarial Networks
Generative adversarial network (GAN) has achieved impressive success on cross-domain generation, but it faces difficulty in cross-modal generation due to the lack of a common distribution between heterogeneous data. Most existing methods of conditional based cross-modal GANs adopt the strategy of one-directional transfer and have achieved preliminary success on text-to-image transfer. Instead o...
متن کاملLatent Constraints: Learning to Generate Conditionally from Unconditional Generative Models
Deep generative neural networks have proven effective at both conditional and unconditional modeling of complex data distributions. Conditional generation enables interactive control, but creating new controls often requires expensive retraining. In this paper, we develop a method to condition generation without retraining the model. By post-hoc learning latent constraints, value functions that...
متن کاملAnalysis and Comparison of Load Flow Methods for Distribution Networks Considering Distributed Generation
Conventional passive distribution networks are changing to modern active distribution networks which are not radial. Conventional load flow methods should be modified for new distribution networks analysis. In modern distribution networks distributed generation (DG) units are embedded with conventional and/or renewable resources. DG units are generally modeled as PV or PQ nodes which inject ...
متن کاملCompact and Efficient Presentation Conversion Routines
Presentation conversion is a key operation in any development environment for distributed applications, such as Corba, Java-RMI, DCE or ASN.1-based environments. It is also well-known performance bottleneck in high-speed network communication. Presentation conversion code is usually generated by an automatic code generation tool referred to as stub compiler. The quality of the code generated by...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1603.06744 شماره
صفحات -
تاریخ انتشار 2016